Pomerania Province
The toddler who survived a 54-degree body temperature
Humans aren't built for the cold, but have survived frigid temperatures in some amazing cases. Breakthroughs, discoveries, and DIY tips sent six days a week. Winter is not for the faint of heart. In New York City, skyscrapers turn Manhattan into a series of freezing wind tunnels. In Sapporo, Japan, the snowfall is almost 200 inches each winter. Even so, humans have developed plenty of clever ways to wait out the cold. But what would happen if instead of bundling up inside with a hot chocolate, you were left in the frigid cold--just how cold can humans get and recover?
- North America > United States > New York (0.25)
- Asia > Japan > Hokkaidō > Hokkaidō Prefecture > Sapporo (0.24)
- Europe > Russia > Central Federal District > Moscow Oblast > Moscow (0.04)
- (3 more...)
- Health & Medicine > Therapeutic Area (0.73)
- Health & Medicine > Diagnostic Medicine > Vital Signs (0.45)
- Information Technology > Communications > Mobile (0.42)
- Information Technology > Artificial Intelligence (0.35)
Model-Based Ranking of Source Languages for Zero-Shot Cross-Lingual Transfer
Ebrahimi, Abteen, Wiemerslage, Adam, von der Wense, Katharina
We present NN-Rank, an algorithm for ranking source languages for cross-lingual transfer, which leverages hidden representations from multilingual models and unlabeled target-language data. We experiment with two pretrained multilingual models and two tasks: part-of-speech tagging (POS) and named entity recognition (NER). We consider 51 source languages and evaluate on 56 and 72 target languages for POS and NER, respectively. When using in-domain data, NN-Rank beats state-of-the-art baselines that leverage lexical and linguistic features, with average improvements of up to 35.56 NDCG for POS and 18.14 NDCG for NER. As prior approaches can fall back to language-level features if target language data is not available, we show that NN-Rank remains competitive using only the Bible, an out-of-domain corpus available for a large number of languages. Ablations on the amount of unlabeled target data show that, for subsets consisting of as few as 25 examples, NN-Rank produces high-quality rankings which achieve 92.8% of the NDCG achieved using all available target data for ranking.
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Europe > Hungary > Csongrád-Csanád County > Szeged (0.04)
- Europe > Poland > Pomerania Province > Gdańsk (0.04)
- (15 more...)
Lecture Notes on Verifying Graph Neural Networks
In these lecture notes, we first recall the connection between graph neural networks, Weisfeiler-Lehman tests and logics such as first-order logic and graded modal logic. We then present a modal logic in which counting modalities appear in linear inequalities in order to solve verification tasks on graph neural networks. We describe an algorithm for the satisfiability problem of that logic. It is inspired from the tableau method of vanilla modal logic, extended with reasoning in quantifier-free fragment Boolean algebra with Presburger arithmetic.
Integrating Domain Knowledge into Process Discovery Using Large Language Models
Norouzifar, Ali, Kourani, Humam, Dees, Marcus, van der Aalst, Wil
Process discovery aims to derive process models from event logs, providing insights into operational behavior and forming a foundation for conformance checking and process improvement. However, models derived solely from event data may not accurately reflect the real process, as event logs are often incomplete or affected by noise, and domain knowledge, an important complementary resource, is typically disregarded. As a result, the discovered models may lack reliability for downstream tasks. We propose an interactive framework that incorporates domain knowledge, expressed in natural language, into the process discovery pipeline using Large Language Models (LLMs). Our approach leverages LLMs to extract declarative rules from textual descriptions provided by domain experts. These rules are used to guide the IMr discovery algorithm, which recursively constructs process models by combining insights from both the event log and the extracted rules, helping to avoid problematic process structures that contradict domain knowledge. The framework coordinates interactions among the LLM, domain experts, and a set of backend services. We present a fully implemented tool that supports this workflow and conduct an extensive evaluation of multiple LLMs and prompt engineering strategies. Our empirical study includes a case study based on a real-life event log with the involvement of domain experts, who assessed the usability and effectiveness of the framework.
- Europe > Austria > Vienna (0.14)
- Europe > Portugal (0.04)
- North America > United States > New Mexico > Santa Fe County > Santa Fe (0.04)
- (15 more...)
- Workflow (1.00)
- Research Report (1.00)
- Overview (1.00)
When Models Lie, We Learn: Multilingual Span-Level Hallucination Detection with PsiloQA
Rykov, Elisei, Petrushina, Kseniia, Savkin, Maksim, Olisov, Valerii, Vazhentsev, Artem, Titova, Kseniia, Panchenko, Alexander, Konovalov, Vasily, Belikova, Julia
Hallucination detection remains a fundamental challenge for the safe and reliable deployment of large language models (LLMs), especially in applications requiring factual accuracy. Existing hallucination benchmarks often operate at the sequence level and are limited to English, lacking the fine-grained, multilingual supervision needed for a comprehensive evaluation. In this work, we introduce PsiloQA, a large-scale, multilingual dataset annotated with span-level hallucinations across 14 languages. PsiloQA is constructed through an automated three-stage pipeline: generating question-answer pairs from Wikipedia using GPT-4o, eliciting potentially hallucinated answers from diverse LLMs in a no-context setting, and automatically annotating hallucinated spans using GPT-4o by comparing against golden answers and retrieved context. We evaluate a wide range of hallucination detection methods -- including uncertainty quantification, LLM-based tagging, and fine-tuned encoder models -- and show that encoder-based models achieve the strongest performance across languages. Furthermore, PsiloQA demonstrates effective cross-lingual generalization and supports robust knowledge transfer to other benchmarks, all while being significantly more cost-efficient than human-annotated datasets. Our dataset and results advance the development of scalable, fine-grained hallucination detection in multilingual settings.
Ultra-Fast Language Generation via Discrete Diffusion Divergence Instruct
Zheng, Haoyang, Liu, Xinyang, Kong, Cindy Xiangrui, Jiang, Nan, Hu, Zheyuan, Luo, Weijian, Deng, Wei, Lin, Guang
Fast and high-quality language generation is the holy grail that people pursue in the age of AI. In this work, we introduce Discrete Diffusion Divergence Instruct (DiDi-Instruct), a training-based method that initializes from a pre-trained (masked) discrete diffusion language model (dLLM) and distills a few-step student for fast generation. The resulting DiDi-Instruct model achieves comparable or superior performance to its dLLM teacher and the GPT-2 baseline while enabling up to 64$\times$ acceleration. The theoretical foundation of DiDi-Instruct is a novel framework based on integral KL-divergence minimization, which yields a practical training algorithm. We further introduce grouped reward normalization, intermediate-state matching, and the reward-guided ancestral sampler that significantly improve training stability, model coverage, and inference quality. On OpenWebText, DiDi-Instruct achieves perplexity from 62.2 (8 NFEs) to 18.4 (128 NFEs), which outperforms prior accelerated dLLMs and GPT-2 baseline. These gains come with a negligible entropy loss (around $1\%$) and reduce additional training wall-clock time by more than $20\times$ compared to competing dLLM distillation methods. We further validate the robustness and effectiveness of DiDi-Instruct through extensive ablation studies, model scaling, and the generation of discrete protein sequences. In conclusion, DiDi-Instruct is an efficient yet effective distillation method, enabling language generation in the blink of an eye. We will release both code and models at github.com/haoyangzheng-ai/didi-instruct.
- Asia > Russia (0.93)
- Asia > India (0.14)
- North America > United States > Virginia (0.04)
- (33 more...)
- Research Report > New Finding (1.00)
- Personal > Interview (0.92)
Dual Branch VideoMamba with Gated Class Token Fusion for Violence Detection
Senadeera, Damith Chamalke, Yang, Xiaoyun, Li, Shibo, Awais, Muhammad, Kollias, Dimitrios, Slabaugh, Gregory
The rapid proliferation of surveillance cameras has increased the demand for automated violence detection. While CNNs and Transformers have shown success in extracting spatio-temporal features, they struggle with long-term dependencies and computational efficiency. W e propose Dual Branch VideoMamba with Gated Class T oken Fusion (GCTF), an efficient architecture combining a dual-branch design and a state-space model (SSM) backbone where one branch captures spatial features, while the other focuses on temporal dynamics. The model performs continuous fusion via a gating mechanism between the branches to enhance the model's ability to detect violent activities even in challenging surveillance scenarios. W e also present a new benchmark by merging RWF-2000, RLVS, SURV and VioPeru datasets in video violence detection, ensuring strict separation between training and testing sets. Experimental results demonstrate that our model achieves state-of-the-art performance on this benchmark and also on DVD dataset which is another novel dataset on video violence detection, offering an optimal balance between accuracy and computational efficiency, demonstrating the promise of SSMs for scalable, near real-time surveillance violence detection.
- North America > United States (0.14)
- Europe > United Kingdom > England > Greater London > London (0.04)
- South America > Peru (0.04)
- (2 more...)
Analysing Python Machine Learning Notebooks with Moose
Mignard, Marius, Costiou, Steven, Anquetil, Nicolas, Etien, Anne
Machine Learning (ML) code, particularly within notebooks, often exhibits lower quality compared to traditional software. Bad practices arise at three distinct levels: general Python coding conventions, the organizational structure of the notebook itself, and ML-specific aspects such as reproducibility and correct API usage. However, existing analysis tools typically focus on only one of these levels and struggle to capture ML-specific semantics, limiting their ability to detect issues. This paper introduces Vespucci Linter, a static analysis tool with multi-level capabilities, built on Moose and designed to address this challenge. Leveraging a metamodeling approach that unifies the notebook's structural elements with Python code entities, our linter enables a more contextualized analysis to identify issues across all three levels. We implemented 22 linting rules derived from the literature and applied our tool to a corpus of 5,000 notebooks from the Kaggle platform. The results reveal violations at all levels, validating the relevance of our multi-level approach and demonstrating Vespucci Linter's potential to improve the quality and reliability of ML development in notebook environments.
- North America > United States > New York > New York County > New York City (0.05)
- Europe > Poland > Pomerania Province > Gdańsk (0.04)
- Europe > France > Hauts-de-France > Nord > Lille (0.04)
Exploring the Impact of Generative Artificial Intelligence on Software Development in the IT Sector: Preliminary Findings on Productivity, Efficiency and Job Security
Bonin, Anton Ludwig, Smolinski, Pawel Robert, Winiarski, Jacek
This study investigates the impact of Generative AI on software development within the IT sector through a mixed-method approach, utilizing a survey developed based on expert interviews. The preliminary results of an ongoing survey offer early insights into how Generative AI reshapes personal productivity, organizational efficiency, adoption, business strategy and job insecurity. The findings reveal that 97% of IT workers use Generative AI tools, mainly ChatGPT. Participants report significant personal productivity gain and perceive organizational efficiency improvements that correlate positively with Generative AI adoption by their organizations (r = .470, p < .05). However, increased organizational adoption of AI strongly correlates with heightened employee job security concerns (r = .549, p < .001). Key adoption challenges include inaccurate outputs (64.2%), regulatory compliance issues (58.2%) and ethical concerns (52.2%). This research offers early empirical insights into Generative AI's economic and organizational implications.
- Europe > Poland > Pomerania Province > Gdańsk (0.05)
- North America > United States > District of Columbia > Washington (0.04)
- North America > United States > New York (0.04)
- (4 more...)
- Questionnaire & Opinion Survey (1.00)
- Overview (0.93)
- Research Report > New Finding (0.69)
- Research Report > Experimental Study (0.46)